103 research outputs found

    Analysis of uncertainty in health care cost-effectiveness studies: an introduction to statistical issues and methods

    Get PDF
    Cost-effectiveness analysis is now an integral part of health technology assessment and addresses the question of whether a new treatment or other health care program offers good value for money. In this paper we introduce the basic framework for decision making with cost-effectiveness data and then review recent developments in statistical methods for analysis of uncertainty when cost-effectiveness estimates are based on observed data from a clinical trial. Although much research has focused on methods for calculating confidence intervals for cost-effectiveness ratios using bootstrapping or Fieller’s method, these calculations can be problematic with a ratio-based statistic where numerator and=or denominator can be zero. We advocate plotting the joint density of cost and effect differences, together with cumulative density plots known as cost-effectiveness acceptability curves (CEACs) to summarize the overall value-for-money of interventions. We also outline the net-benefit formulation of the cost-effectiveness problem and show that it has particular advantages over the standard incremental cost-effectiveness ratio formulation

    Confidence interval estimation for the changepoint of treatment stratification in the presence of a qualitative covariate-treatment interaction

    Get PDF
    The goal in stratified medicine is to administer the \textquotedblbest\textquotedbl treatment to a patient. Not all patients might benefit from the same treatment; the choice of best treatment can depend on certain patient characteristics. In this article, it is assumed that a time-to-event outcome is considered as a patient-relevant outcome and a qualitative interaction between a continuous covariate and treatment exists, ie,~that patients with different values of one specific covariate should be treated differently. We suggest and investigate different methods for confidence interval estimation for the covariate value, where the treatment recommendation should be changed based on data collected in a randomized clinical trial. An adaptation of Fieller's theorem, the delta method, and different bootstrap approaches (normal, percentile-based, wild bootstrap) are investigated and compared in a simulation study. Extensions to multivariable problems are presented and evaluated. We observed appropriate confidence interval coverage following Fieller's theorem irrespective of sample size but at the cost of very wide or even infinite confidence intervals. The delta method and the wild bootstrap approach provided the smallest intervals but inadequate coverage for small to moderate event numbers, also depending on the location of the true changepoint. For the percentile-based bootstrap, wide intervals were observed, and it was slightly conservative regarding coverage, whereas the normal bootstrap did not provide acceptable results for many scenarios. The described methods were also applied to data from a randomized clinical trial comparing two treatments for patients with symptomatic, severe carotid artery stenosis, considering patient's age as predictive marker

    Methods for confidence interval estimation of a ratio parameter with application to location quotients

    Get PDF
    BACKGROUND: The location quotient (LQ) ratio, a measure designed to quantify and benchmark the degree of relative concentration of an activity in the analysis of area localization, has received considerable attention in the geographic and economics literature. This index can also naturally be applied in the context of population health to quantify and compare health outcomes across spatial domains. However, one commonly observed limitation of LQ is its widespread use as only a point estimate without an accompanying confidence interval. METHODS: In this paper we present statistical methods that can be used to construct confidence intervals for location quotients. The delta and Fieller's methods are generic approaches for a ratio parameter and the generalized linear modelling framework is a useful re-parameterization particularly helpful for generating profile-likelihood based confidence intervals for the location quotient. A simulation experiment is carried out to assess the performance of each of the analytic approaches and a health utilization data set is used for illustration. RESULTS: Both the simulation results as well as the findings from the empirical data show that the different analytical methods produce very similar confidence limits for location quotients. When incidence of outcome is not rare and sample sizes are large, the confidence limits are almost indistinguishable. The confidence limits from the generalized linear model approach might be preferable in small sample situations. CONCLUSION: LQ is a useful measure which allows quantification and comparison of health and other outcomes across defined geographical regions. It is a very simple index to compute and has a straightforward interpretation. Reporting this estimate with appropriate confidence limits using methods presented in this paper will make the measure particularly attractive for policy and decision makers

    Local Difference Measures between Complex Networks for Dynamical System Model Evaluation

    Get PDF
    Acknowledgments We thank Reik V. Donner for inspiring suggestions that initialized the work presented herein. Jan H. Feldhoff is credited for providing us with the STARS simulation data and for his contributions to fruitful discussions. Comments by the anonymous reviewers are gratefully acknowledged as they led to substantial improvements of the manuscript.Peer reviewedPublisher PD

    Improving statistical inference on pathogen densities estimated by quantitative molecular methods: malaria gametocytaemia as a case study

    Get PDF
    BACKGROUND: Quantitative molecular methods (QMMs) such as quantitative real-time polymerase chain reaction (q-PCR), reverse-transcriptase PCR (qRT-PCR) and quantitative nucleic acid sequence-based amplification (QT-NASBA) are increasingly used to estimate pathogen density in a variety of clinical and epidemiological contexts. These methods are often classified as semi-quantitative, yet estimates of reliability or sensitivity are seldom reported. Here, a statistical framework is developed for assessing the reliability (uncertainty) of pathogen densities estimated using QMMs and the associated diagnostic sensitivity. The method is illustrated with quantification of Plasmodium falciparum gametocytaemia by QT-NASBA. RESULTS: The reliability of pathogen (e.g. gametocyte) densities, and the accompanying diagnostic sensitivity, estimated by two contrasting statistical calibration techniques, are compared; a traditional method and a mixed model Bayesian approach. The latter accounts for statistical dependence of QMM assays run under identical laboratory protocols and permits structural modelling of experimental measurements, allowing precision to vary with pathogen density. Traditional calibration cannot account for inter-assay variability arising from imperfect QMMs and generates estimates of pathogen density that have poor reliability, are variable among assays and inaccurately reflect diagnostic sensitivity. The Bayesian mixed model approach assimilates information from replica QMM assays, improving reliability and inter-assay homogeneity, providing an accurate appraisal of quantitative and diagnostic performance. CONCLUSIONS: Bayesian mixed model statistical calibration supersedes traditional techniques in the context of QMM-derived estimates of pathogen density, offering the potential to improve substantially the depth and quality of clinical and epidemiological inference for a wide variety of pathogens

    Is Medical Research Informing Professional Practice More Highly Cited? Evidence from AHFS DI Essentials in Drugs.com

    Get PDF
    This is an accepted manuscript of an article published by Springer in Scientometrics on 21/02/2017, available online: https://doi.org/10.1007/s11192-017-2292-3 The accepted version of the publication may differ from the final published version.Citation-based indicators are often used to help evaluate the impact of published medical studies, even though the research has the ultimate goal of improving human wellbeing. One direct way of influencing health outcomes is by guiding physicians and other medical professionals about which drugs to prescribe. A high profile source of this guidance is the AHFS DI Essentials product of the American Society of Health-System Pharmacists, which gives systematic information for drug prescribers. AHFS DI Essentials documents, which are also indexed by Drugs.com, include references to academic studies and the referenced work is therefore helping patients by guiding drug prescribing. This article extracts AHFS DI Essentials documents from Drugs.com and assesses whether articles referenced in these information sheets have their value recognised by higher Scopus citation counts. A comparison of mean log-transformed citation counts between articles that are and are not referenced in AHFS DI Essentials shows that AHFS DI Essentials references are more highly cited than average for the publishing journal. This suggests that medical research influencing drug prescribing is more cited than average

    Genetic properties of feed efficiency parameters in meat-type chickens

    Get PDF
    <p>Abstract</p> <p>Background</p> <p>Feed cost constitutes about 70% of the cost of raising broilers, but the efficiency of feed utilization has not kept up the growth potential of today's broilers. Improvement in feed efficiency would reduce the amount of feed required for growth, the production cost and the amount of nitrogenous waste. We studied residual feed intake (RFI) and feed conversion ratio (FCR) over two age periods to delineate their genetic inter-relationships.</p> <p>Methods</p> <p>We used an animal model combined with Gibb sampling to estimate genetic parameters in a pedigreed random mating broiler control population.</p> <p>Results</p> <p>Heritability of RFI and FCR was 0.42-0.45. Thus selection on RFI was expected to improve feed efficiency and subsequently reduce feed intake (FI). Whereas the genetic correlation between RFI and body weight gain (BWG) at days 28-35 was moderately positive, it was negligible at days 35-42. Therefore, the timing of selection for RFI will influence the expected response. Selection for improved RFI at days 28-35 will reduce FI, but also increase growth rate. However, selection for improved RFI at days 35-42 will reduce FI without any significant change in growth rate. The nature of the pleiotropic relationship between RFI and FCR may be dependent on age, and consequently the molecular factors that govern RFI and FCR may also depend on stage of development, or on the nature of resource allocation of FI above maintenance directed towards protein accretion and fat deposition. The insignificant genetic correlation between RFI and BWG at days 35-42 demonstrates the independence of RFI on the level of production, thereby making it possible to study the molecular, physiological and nutrient digestibility mechanisms underlying RFI without the confounding effects of growth. The heritability estimate of FCR was 0.49 and 0.41 for days 28-35 and days 35-42, respectively.</p> <p>Conclusion</p> <p>Selection for FCR will improve efficiency of feed utilization but because of the genetic dependence of FCR and its components, selection based on FCR will reduce FI and increase growth rate. However, the correlated responses in both FI and BWG cannot be predicted accurately because of the inherent problem of FCR being a ratio trait.</p

    The cost-effectiveness of screening for ovarian cancer: results from the UK Collaborative Trial of Ovarian Cancer Screening (UKCTOCS)

    Get PDF
    Background: To assess the within trial cost-effectiveness of an NHS ovarian cancer screening (OCS) programme using data from UKCTOCS and extrapolate results based on average life expectancy. Methods: Within trial economic evaluation of no screening (C) versus either (1) an annual OCS programme using transvaginal ultrasound (USS) or (2) an annual ovarian cancer multimodal screening programme with serum CA125 interpreted using a risk algorithm (ROCA) and transvaginal ultrasound as a second line test (MMS), plus comparison of lifetime extrapolation of the no screening arm and the MMS programme using both a predictive and a Markov model. Results: Using a CA125-ROCA cost of £20, the within trial results show USS to be strictly dominated by MMS, with the MMS versus C comparison returning an Incremental Cost-Effectiveness ratio (ICER) of £91,452 per life year gained (LYG). If the CA125-ROCA unit cost is reduced to £15 the ICER becomes £77,818 per LYG. Predictive extrapolation over the expected lifetime of the UKCTOCS women returns an ICER of £30,033 per LYG, while Markov modelling produces an ICER of £46,922 per QALY. Conclusions: Analysis suggests that, after accounting for the lead-time required to establish full mortality benefits, a national OCS programme based on the MMS strategy quickly approaches the current NICE thresholds for cost-effectiveness when extrapolated out to lifetime as compared to the within trial ICER estimates. Whether MMS could be recommended on economic grounds would depend on the confirmation and size of the mortality benefit at the end of an ongoing follow-up of the UKCTOCS cohort

    An experimental study of the intrinsic stability of random forest variable importance measures

    Get PDF
    BACKGROUND: The stability of Variable Importance Measures (VIMs) based on random forest has recently received increased attention. Despite the extensive attention on traditional stability of data perturbations or parameter variations, few studies include influences coming from the intrinsic randomness in generating VIMs, i.e. bagging, randomization and permutation. To address these influences, in this paper we introduce a new concept of intrinsic stability of VIMs, which is defined as the self-consistence among feature rankings in repeated runs of VIMs without data perturbations and parameter variations. Two widely used VIMs, i.e., Mean Decrease Accuracy (MDA) and Mean Decrease Gini (MDG) are comprehensively investigated. The motivation of this study is two-fold. First, we empirically verify the prevalence of intrinsic stability of VIMs over many real-world datasets to highlight that the instability of VIMs does not originate exclusively from data perturbations or parameter variations, but also stems from the intrinsic randomness of VIMs. Second, through Spearman and Pearson tests we comprehensively investigate how different factors influence the intrinsic stability. RESULTS: The experiments are carried out on 19 benchmark datasets with diverse characteristics, including 10 high-dimensional and small-sample gene expression datasets. Experimental results demonstrate the prevalence of intrinsic stability of VIMs. Spearman and Pearson tests on the correlations between intrinsic stability and different factors show that #feature (number of features) and #sample (size of sample) have a coupling effect on the intrinsic stability. The synthetic indictor, #feature/#sample, shows both negative monotonic correlation and negative linear correlation with the intrinsic stability, while OOB accuracy has monotonic correlations with intrinsic stability. This indicates that high-dimensional, small-sample and high complexity datasets may suffer more from intrinsic instability of VIMs. Furthermore, with respect to parameter settings of random forest, a large number of trees is preferred. No significant correlations can be seen between intrinsic stability and other factors. Finally, the magnitude of intrinsic stability is always smaller than that of traditional stability. CONCLUSION: First, the prevalence of intrinsic stability of VIMs demonstrates that the instability of VIMs not only comes from data perturbations or parameter variations, but also stems from the intrinsic randomness of VIMs. This finding gives a better understanding of VIM stability, and may help reduce the instability of VIMs. Second, by investigating the potential factors of intrinsic stability, users would be more aware of the risks and hence more careful when using VIMs, especially on high-dimensional, small-sample and high complexity datasets
    • …
    corecore